Estimation of Covariance Matrices under Sparsity Constraints
نویسندگان
چکیده
Discussion of “Minimax Estimation of Large Covariance Matrices under L1-Norm” by Tony Cai and Harrison Zhou. To appear in Statistica Sinica. Introduction. Estimation of covariance matrices in various norms is a critical issue that finds applications in a wide range of statistical problems, and especially in principal component analysis. It is well known that, without further assumptions, the empirical covariance matrix Σ is the best possible estimator in many ways, and in particular in a minimax sense. However, it is also well known that Σ is not an accurate estimator when the dimension p of the observations is high. The minimax analysis carried out by Tony Cai and Harry Zhou ([CZ] in what follows) guarantees that for several classes of matrices with reasonable structure (sparse or banded matrices), the fully data-driven thresholding estimator achieves the best possible rates when p is much larger than the sample size n. This is done, in particular, by proving minimax lower bounds that ensure that no estimator can perform better than the hard thresholding estimator, uniformly over the sparsity classes Gq for each 0 ≤ q < 1. This result has a flavor of universality in the sense that one and the same estimator is minimax optimal for several classes of matrices. Our comments focus on the sparsity classes of matrices. (a) Optimal rates. Optimal rates are obtained in [CZ] under the assumption that the dimension is very high: p ≥ nν, ν > 1. Thus, the case of dimensions smaller than n, or even p ≈ n, is excluded. This seems to be due to the technique of proving the lower bound (Theorem 2 in [CZ]). Indeed, by a different technique, we show that the lower bound holds without this assumption, cf. Theorem 1 below. Furthermore, in general, our lower rate ψ(1) is different from that obtained in [CZ] and has ingredients similar to the optimal rate for the Gaussian sequence model. We conjecture that it is optimal for all admissible configurations of n, p, and sparsity parameters. (b) Frobenius norm and global sparsity. We argue that the Frobenius norm is naturally adapted to the structure of the problem, at least for Gaussian observations, and we derive optimal rates under the Frobenius risk and global sparsity assumption. Department of Operations Research and Financial Engineering, Princeton University, Princeton, NJ 08544, USA. Supported in part by NSF grants DMS-0906424 and CAREER-DMS-1053987. Laboratoire de Statistique, CREST-ENSAE, 3, av. Pierre Larousse, F-92240 Malakoff Cedex, France. Supported in part by ANR under the grant “Parcimonie”.
منابع مشابه
Sparse estimation of high-dimensional correlation matrices
Estimating covariations of variables for high dimensional data is important for understanding their relations. Recent years have seen several attempts to estimate covariance matrices with sparsity constraints. A new convex optimization formulation for estimating correlation matrices, which are scale invariant, is proposed as opposed to covariance matrices. The constrained optimization problem i...
متن کاملMinimax Estimation of Bandable Precision Matrices
The inverse covariance matrix provides considerable insight for understanding statistical models in the multivariate setting. In particular, when the distribution over variables is assumed to be multivariate normal, the sparsity pattern in the inverse covariance matrix, commonly referred to as the precision matrix, corresponds to the adjacency matrix representation of the Gauss-Markov graph, wh...
متن کاملPAGE: Robust Pattern Guided Estimation of Large Covariance Matrix
We study the problem of estimating large covariance matrices under two types of structural assumptions: (i) The covariance matrix is the summation of a low rank matrix and a sparse matrix, and we have some prior information on the sparsity pattern of the sparse matrix; (ii) The data follow a transelliptical distribution. The former structure regulates the parameter space and has its roots in di...
متن کاملCovariance Matrix Estimation for Stationary Time Series
We obtain a sharp convergence rate for banded covariance matrix estimates of stationary processes. A precise order of magnitude is derived for spectral radius of sample covariance matrices. We also consider a thresholded covariance matrix estimator that can better characterize sparsity if the true covariance matrix is sparse. As our main tool, we implement Toeplitz [Math. Ann. 70 (1911) 351–376...
متن کامل0 Sparse Inverse Covariance Estimation
Recently, there has been focus on penalized loglikelihood covariance estimation for sparse inverse covariance (precision) matrices. The penalty is responsible for inducing sparsity, and a very common choice is the convex l1 norm. However, the best estimator performance is not always achieved with this penalty. The most natural sparsity promoting “norm” is the non-convex l0 penalty but its lack ...
متن کامل